Learning to Solve Complex Planning Problems: Finding Useful Auxiliary Problems
نویسندگان
چکیده
Learning from past experience allows a problem solver to increase its solvability horizon from simple to complex problems. For planners, learning involves a training phase during which knowledge is extracted from simple problems. But how are these simple problems constructed? All current learning and problem solving systems require the user to provide the training set. However it is rarely easy to identify problems that are both simple and useful for learning, especially in complex applications. In this paper, we present our initial research towards the automated or semiautomated identification of these simple problems. From a difficult problem and a corresponding partially completed search episode, we extract auxiliary problems with which to train the learner. We motivate this overlooked issue, describe our approach, and illustrate it with examples. Introduction and Problem Formulation Researchers in Machine Learning and in Planning have developed several systems that use simple planning problems to learn how to solve more difficult problems (among several others, (Laird, Rosenbloom, & Newell 1986; DeJong & Mooney 1986; Minton 1988; Veloso & Carbonell 1993; Borrajo & Veloso 1994)). However, all current systems require that the simpler problems be provided by the user. This requirement is a gap in automated learning that we propose to fill. Using a previously untried approach, we are developing a system that will find auxiliary problems that are likely to be useful in learning how to solve a difficult problem. A planning problem can be considered difficult for a particular planner if it cannot solve the problem with a reasonable amount of effort. Learning how to solve difficult problems is a three-step process: First, the learner must find some simpler, i.e. more directly solvable, problems that are somehow related to the original problem. These problems are called auxiliary problems (Polya 1945). This first step is the one which we address here. Second, the learner must take these auxiliary problems and learn by solving them. Since they are simpler than the original problem, the planner should be able to solve the auxiliary problems with relatively little effort. Third, the learner must use the knowledge gained by solving the auxiliary problems to find a solution to the original problem. This process is successful if the time to carry out all three steps is significantly less than the time it would have taken to solve the original problem without the benefit of learning. Current learning systems work towards executing the second and third steps of the above process while finessing the first. Training problems from which to learn are usually generated randomly, entered by the user, or defined by a set of rules according to some preset measure of simplicity, such as number of goals, which is not necessarily accurate (Minton 1988; Etzioni 1993; Knoblock 1994; Bhansali 1991; Veloso 1992; Katukam & Kambhampati 1994). The goal of our research is to provide a general method for finding auxiliary problems that are most likely to help us learn how to approach a difficult problem. To be useful, these auxiliary problems must be solvable with relatively little effort on the part of the planner; otherwise they are no more approachable than the original problem. Yet the auxiliary problems must also retain some complexities, if only on a smaller, more learnable scale. If we simplify the problem too much we will not learn anything relevant to the original problem. Finding auxiliary problems that are neither too complex nor too simple is a difficult task. We believe that we will be able to accomplish this task because we are considering some available information that previous attempts at this task have ignored. Some of the few attempts at decomposing a problem into simpler problems are based on capturing different abstraction levels or problem spaces in the domain definition (Knoblock 1994; Rosenbloom, Newell, & Laird 1990; Sacerdoti 1974). These approaches have one thing in common: a static analysis of the domain and problem, either automated or done by the user. But simply looking at the problem may not provide any information as to what is causing the planner to have difficulty. We propose to examine not only the original problem and domain definition, but also the planner’s unsuccessful solution attempt. The partially completed search space from the original problem will help us discover why the problem is so hard for a particular planner. In this way, we will be able to determine what aspects of the original problem we most need to learn.
منابع مشابه
Learning to Plan in Complex Stochastic Domains
Probabilistic planning offers a powerful framework for general problem solving. Historically, probabilistic planning algorithms have contributed to a variety of critical application areas and technologies, including conservation biology [42], self-driving cars [48, 37], and space exploration [14, 6, 20]. However, optimal planning is known to be P-Complete with respect to the size of the state-a...
متن کاملPERFORMANCE COMPARISON OF CBO AND ECBO FOR LOCATION FINDING PROBLEMS
The p-median problem is one of the discrete optimization problem in location theory which aims to satisfy total demand with minimum cost. A high-level algorithmic approach can be specialized to solve optimization problem. In recent years, meta-heuristic methods have been applied to support the solution of Combinatorial Optimization Problems (COP). Collision Bodies Optimization algorithm (CBO) a...
متن کاملActive model learning and diverse action sampling for task and motion planning
The objective of this work is to augment the basic abilities of a robot by learning to use new sensorimotor primitives to enable the solution of complex long-horizon problems. Solving long-horizon problems in complex domains requires flexible generative planning that can combine primitive abilities in novel combinations to solve problems as they arise in the world. In order to plan to combine p...
متن کاملSimulation Methods for Uncertain Decision-Theoretic Planning
Experience based reinforcement learning (RL) systems are known to be useful for dealing with domains that are a priori unknown. We believe that experience based methods may also be useful when the model is uncertain (or even completely known). In this case experience is gained by simulating the uncertain model. This paper explores a simple way to allow experience based RL systems to cope with u...
متن کاملToward a Model of Intelligence as an Economy of Idiots
A market-based algorithm is presented which autonomously apportions complex tasks to multiple cooperating agents giving each agent the motivation of improving performance of the whole system. A speciic model, called \The Hayek Machine" is proposed and tested on a simulated Blocks World (BW) planning problem. Hayek learns to solve more complex BW problems than any previous learning algorithm. Gi...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1994